45 research outputs found

    A hashtag worth a thousand words: Discursive strategies around #JeNeSuisPasCharlie after the 2015 Charlie Hebdo shooting

    Get PDF
    Following a shooting attack by two self-proclaimed Islamist gunmen at the offices of French satirical weekly Charlie Hebdo on 7 January 2015, there emerged the hashtag #JeSuisCharlie on Twitter as an expression of solidarity and support for the magazine’s right to free speech. Almost simultaneously, however, there was also #JeNeSuisPasCharlie explicitly countering the former, affirmative hashtag. Based on a multimethod analysis of 74,047 tweets containing #JeNeSuisPasCharlie posted between 7 and 11 January, this article reveals that users of the hashtag under study employed various discursive strategies and tactics to challenge the mainstream framing of the shooting as the universal value of freedom of expression being threatened by religious extremism, while protecting themselves from the risk of being viewed as disrespecting victims or endorsing the violence committed. The significance of this study is twofold. First, it extends the literature on strategic speech acts by examining how such acts take place in a social media context. Second, it highlights the need for a multidimensional and reflective methodology when dealing with data mined from social media

    Twitter as health information source : exploring the parameters affecting dementia-related tweets

    Get PDF
    Unlike other media, research on the credibility of information present on social media is limited. This limitation is even more pronounced in the case of healthcare, including dementia-related information. The purpose of this study was to identify user groups that show high bot-like behavior and profile features that deviation from typical human behavior. We collected 16,691 tweets about dementia posted over the course of a month by 8400 users. We applied inductive coding to categorize users. The BotOrNot? API was used to compute a bot score. This work provides insight into relations between user features and a bot score. We performed analysis techniques such as Kruskal-Wallis, stepwise multiple variable regression, user tweet frequency analysis and content analysis on the data. These were further evaluated for the most frequently referenced URLs in the tweets and most active users in terms of tweet frequency. Initial results indicated that the majority of users are regular users and not bots. Regression analysis revealed a clear relationship between different features. Independent variables in the user profiles such as geo_data and favourites_count, correlated with the final bot score. Similarly, content analysis of the tweets showed that the word features of bot profiles have an overall smaller percentage of words compared to regular profiles. Although this analysis is promising, it needs further enhancements

    A meta-analysis of state-of-the-art electoral prediction from Twitter data

    Full text link
    Electoral prediction from Twitter data is an appealing research topic. It seems relatively straightforward and the prevailing view is overly optimistic. This is problematic because while simple approaches are assumed to be good enough, core problems are not addressed. Thus, this paper aims to (1) provide a balanced and critical review of the state of the art; (2) cast light on the presume predictive power of Twitter data; and (3) depict a roadmap to push forward the field. Hence, a scheme to characterize Twitter prediction methods is proposed. It covers every aspect from data collection to performance evaluation, through data processing and vote inference. Using that scheme, prior research is analyzed and organized to explain the main approaches taken up to date but also their weaknesses. This is the first meta-analysis of the whole body of research regarding electoral prediction from Twitter data. It reveals that its presumed predictive power regarding electoral prediction has been rather exaggerated: although social media may provide a glimpse on electoral outcomes current research does not provide strong evidence to support it can replace traditional polls. Finally, future lines of research along with a set of requirements they must fulfill are provided.Comment: 19 pages, 3 table

    Noise audits improve moral foundation classification

    No full text
    Morality plays an important role in culture, identity, and emotion. Recent advances in natural language processing have shown that it is possible to classify moral values expressed in text at scale. Morality classification relies on human annotators to label the moral expressions in text, which provides training data to achieve state-of-the-art performance. However, these annotations are inherently subjective and some of the instances are hard to classify, resulting in noisy annotations due to error or lack of agreement. The presence of noise in training data harms the classifier's ability to accurately recognize moral foundations from text. We propose two metrics to audit the noise of annotations. The first metric is entropy of instance labels, which is a proxy measure of annotator disagreement about how the instance should be labeled. The second metric is the silhouette coefficient of a label assigned by an annotator to an instance. This metric leverages the idea that instances with the same label should have similar latent representations, and deviations from collective judgments are indicative of errors. Our experiments on three widely used moral foundations datasets show that removing noisy annotations based on the proposed metrics improves classification performance.11Our code can be found at: https://github.com/negar-mokhberian/noise-audits

    Scalable Vaccine Distribution in Large Graphs given Uncertain Data

    No full text

    Evaluation without ground truth in social media research

    No full text
    corecore